44 research outputs found
system architecture for approximate query processing
Decision making is an activity that addresses the problem of extracting knowledge and information from data stored in data warehouses, in order to improve the business processes of information systems. Usually, decision making is based on On-Line Analytical Processing, data mining, or approximate query processing. In the last case, answers to analytical queries are provided in a fast manner, although affected with a small percentage of error. In the paper, we present the architecture of an approximate query answering system. Then, we illustrate our ADAP (Analytical Data Profile) system, which is based on an engine able to provide fast responses to the main statistical functions by using orthogonal polynomials series to approximate the data distribution of multidimensional relations. Moreover, several experimental results to measure the approximation error are shown and the response-time to analytical queries is reported.</p
Design process for Big Data Warehouses
Big Data Warehouses differ substantially from traditional data warehouses in that their schema should be based on novel logical models that allow more flexibility than that the relational model does. Furthermore, their design methodology also requires new principles, such as automation and agile techniques, in order to gain both a fast realization and reaction to changes in business requirements. In the paper, we discuss the new features of big data warehouses and present an innovative design methodology based on the key-value model
Metadata for Approximate Query Answering Systems
In business intelligence systems, data warehouse metadata management and representation are getting more and more attention
by vendors and designers. The standard language for the data warehouse metadata representation is the Common Warehouse
Metamodel. However, business intelligence systems include also approximate query answering systems, since these software tools
provide fast responses for decisionmaking on the basis of approximate query processing. Currently, the standard meta-model does
not allow to represent the metadata needed by approximate query answering systems. In this paper, we propose an extension of
the standard metamodel, in order to define the metadata to be used in online approximate analytical processing. These metadata
have been successfully adopted in ADAP, a web-based approximate query answering system that creates and uses statistical data
profiles
Academic data warehouse design using a hybrid methodology
In the last years, data warehousing has got attention from Universities which are now adopting business intelligence solutions in order to analyze crucial aspects of the academic context. In this paper, we present the architecture of a Business Intelligence system for academic organizations. Then, we illustrate the design process of the data warehouse devoted to the analysis of the main factors affecting the importance and the quality level of every University, such as the evaluation of the Research and the Didactics. The design process we describe is based on a hybrid methodology that is largely automatic and relies on an ontological approach for the integration of the different data sources
Research Data Mart in an Academic System
Data warehousing is an activity that is getting more and more attention in several contexts. Also Universities are adopting data warehousing solutions for business intelligence purpose. In these contexts, there are specific aspects to be considered, such as the Didactics and the Research evaluation. Indeed, these are the main factors affecting the importance and the quality level of every University. In this paper, we present the architecture of a Business Intelligence system in an academic organization and we illustrate the design of a data mart devoted to the evaluation of the Research activities
A Framework for Evaluating Design Methodologies for Big Data Warehouses: Measurement of the Design Process
This article describes how the evaluation of modern data warehouses considers new solutions adopted for facing the radical changes caused by the necessity of reducing the storage volume, while increasing the velocity in multidimensional design and data elaboration, even in presence of unstructured data that are useful for providing qualitative information. The aim is to set up a framework for the evaluation of the physical and methodological characteristics of a data warehouse, realized by considering the factors that affect the data warehouse’s lifecycle when taking into account the Big Data issues (Volume, Velocity, Variety, Value, and Veracity). The contribution is the definition of a set of criteria for classifying Big Data Warehouses on the basis of their methodological characteristics. Based on these criteria, the authors defined a set of metrics for measuring the quality of Big Data Warehouses in reference to the design specifications. They show through a case study how the proposed metrics are able to check the eligibility of methodologies falling in different classes in the Big Data context
Benchmark for approximate query answering systems
The standard benchmark for Decision Support Systems is TPC-H, which is composed of a database, a workload, and a set of metrics for the performance evaluation. However, TPC-H does not include a methodology for the benchmark of Approximate Query Answering Systems, or the software tools used to obtain fast answers to analytical queries in the decision making process. In the paper, the authors present a methodology to evaluate and compare Approximate Query Answering Systems. To this aim, a methodology that extends the standard TPC-H and a set of new metrics that take into account the specific features of these systems are proposed. Experimental results show the application of these metrics to two systems based on the data analytic approximation by orthonormal series